Constructing Deterministic Finite-State Automata in Recurrent Neural Networksy

نویسنده

  • Christian W. Omlin
چکیده

Recurrent neural networks that are trained to behave like deterministic nite-state automata (DFA's) can show deteriorating performance when tested on long strings. This deteriorating performance can be attributed to the instability of the internal representation of the learned DFA states. The use of a sigmoidal discriminant function together with the recurrent structure contribute to this instability. We prove that a simple constructive algorithm can construct second-order recurrent neural networks with a sparse interconnection topology and sigmoidal discriminant function such that the internal DFA state representations are stable, i.e. the constructed network correctly classiies strings of arbitrary length. The algorithm is based on encoding strengths of weights directly into the neural network. We derive a relationship between the weight strength and the number of DFA states for robust string classiication. For a DFA with n states and m input alphabet symbols, the constructive algorithm generates a \pro-grammed" neural network with O(n) neurons and O(mn) weights. We compare our algorithm to other methods proposed in the literature.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Evolutionary Training of Hybrid Systems of Recurrent Neural Networks and Hidden Markov Models

We present a hybrid architecture of recurrent neural networks (RNNs) inspired by hidden Markov models (HMMs). We train the hybrid architecture using genetic algorithms to learn and represent dynamical systems. We train the hybrid architecture on a set of deterministic finite-state automata strings and observe the generalization performance of the hybrid architecture when presented with a new se...

متن کامل

Efficient encodings of finite automata in discrete-time recurrent neural networks∗

A number of researchers have used discretetime recurrent neural nets (DTRNN) to learn finite-state machines (FSM) from samples of input and output strings; trained DTRNN usually show FSM behaviour for strings up to a certain length, but not beyond; this is usually called instability. Other authors have shown that DTRNN may actually behave as FSM for strings of any length and have devised strate...

متن کامل

Recent Advances of Grammatical Inference

In this paper, we provide a survey of recent advances in the field “Grammatical Inference” with a particular emphasis on the results concerning the learnability of target classes represented by deterministic finite automata, context-free grammars, hidden Markov models, stochastic contextfree grammars, simple recurrent neural networks, and case-based representations.

متن کامل

Recurrent Neural Networks Learn Deterministic Representations of Fuzzy Finite-State Automata

The paradigm of deterministic nite-state automata (DFAs) and their corresponding regular languages have been shown to be very useful for addressing fundamental issues in recurrent neural networks. The issues that have been addressed include knowledge representation, extraction, and reenement as well development of advanced learning algorithms. Recurrent neural networks are also very promising t...

متن کامل

Representation of Fuzzy Finite State

Based on previous work on encoding deterministic nite-state automata (DFAs) in discrete-time, second-order recurrent neural networks with sigmoidal discriminant functions, we propose an algorithm that constructs an augmented recurrent neural network that encodes fuzzy nite-state automata (FFAs). Given an arbitrary FFA, we apply an algorithm which transforms the FFA into an equivalent determinis...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1996